专利摘要:
The present invention relates to a method for controlling subtitle presentation modes. It relates in particular to a process for creating a stream or subtitle file integrating the formation of subtitle blocks according to a set of subtitle presentation modes. It also relates to a method of presenting a stream or subtitle file on the terminal of a user integrating the selection of a subtitle presentation mode, reading a block of subtitles associated with said presentation mode in the stream or file, and the presentation of said block of subtitles on the user's terminal.
公开号:FR3025925A1
申请号:FR1458793
申请日:2014-09-17
公开日:2016-03-18
发明作者:Xavier Besnard;Michel Jouan
申请人:France Brevets SAS;
IPC主号:
专利说明:

[0001] FIELD OF THE INVENTION [0001] The present invention relates to the field of video subtitling, and in particular of real-time captioning. It refers to both the generation of subtitles and the presentation of them.
[0002] PRIOR ART [0002] The application of subtitles is an important element in the production of video streams. Indeed, these subtitles allow a user to have a better understanding of the content of the video, especially the spoken words, when they are little or not understandable. This is particularly useful when a program is broadcast in a language not known to the user, or when the auditory perception of the language is disturbed by poor transmission, poorly articulated speech or background noise. Finally, subtitles provide hearing-impaired people with access to program understanding. The application of subtitles can be carried out mainly in two ways: offline or real-time (so-called subtitles live in the current language of the domain). The application of delayed subtitles is done on a support prepared in advance. This type of subtitling can be found, for example, in DVDs, Blu-ray discs or broadcast programs. This offline application does not have any particular time constraint. Thus, it is possible in this case to manually insert subtitles in the video stream, an operator verifying that the subtitles are perfectly synchronized with the video, while having a visual appearance perceived as pleasant by the user. Although subjective, this notion of pleasant perception of subtitles can be reduced to objective elements, for example to obtain subtitles which are displayed at a moderate speed, or which maintain a fixed position relative to the screen. It is also possible to use audio analysis methods which can be expensive in computing time, in order to synchronize the subtitles and the audio / video as well as possible. In the case of live subtitles, it is however not possible to use these techniques. Indeed, in this case the video content produced is immediately transmitted. Live subtitle production is generally performed by an operator producing live subtitles and sending them to the stream. This operation produces an inevitable delay between the moment the video is transmitted and the corresponding time or subtitle is produced. This delay is even more important when the operator's task involves translation from one language to another for the production of the subtitles. [0005] The approach generally used to process live subtitles is to transmit each subtitle word as soon as it is available, so as not to add additional delay. This approach, however, brings additional disadvantages for the user: in addition to an inevitable delay, the words arrive one after the other, not always forming a coherent whole. In addition, when the subtitles are formed from a teletext source, the line change may cause the words to be shifted upwards. This upward shift, or "shift up" according to the English terminology, consists, when a new line of subtitles starts, to shift all the subtitle lines of a line upwards. and delete the oldest. This effect may be particularly disturbing for the viewer, because if a word he was reading is shifted upward, he will have to make an extra effort to follow that word while retaining the overall meaning of the subtitle. [0006] Live subtitles are thus often perceived by users as unpleasant and of poor quality. The presentation of subtitle words has been listed as one of the main causes of poor perception of live subtitles by users by Ofcom (Office of 30 Communications), The quality of live subtitling, pp. 29-30, http: //stakeholders.ofcom.orq.u/inaries/consultations/subtitlinq/summar/su bffiling.pdf. According to this study, users seem to prefer the subtitles presented by blocks. However, the definition of a "block" remains wide, and some users may prefer subtitles presented line by line, while others prefer subtitles presented sentence by sentence, or again word by word. The known methods of the state of the art consist in producing or updating a page of subtitles in a video encoder, and decode 5 and then display this page or update page. Subtitles can be encoded as images, such as in the DVBSUB (Digital Video Braodcasting Subtitles) standard. In this case, the subtitle style (color, font size ...) is totally defined by the image. They can also be encoded as text characters, as for example in the various standards based on W3C TTML (according to the acronym World Wide Web Consortium Timed Text Markup Language, or Consortium Web on Width of the World Extensible Text Language Temporized Text In this case, the subtitles are stored as 15 characters forming words as in a text file, and possibly header information specifying a style of presentation to be applied. The presentation style then contains information such as the size of characters to apply, the font, etc ... For example, the presentation styles available in the standard EBU-TT (from the acronym 20 European Broadcasting Union-Timed Text, or European Union Broadcasting - Text Timed), from W3C TTML, are described in the EBU-EBU document, EBU-TT Part 1, Subtitling format definition, (https://tech.ebu.ch/ docs / tech / tech3350.pdf). [0009] US Pat. No. 869,5048 describes a method for transcribing subtitles for the deaf and hard of hearing from a first format to a second format independent of the platform. This second format may notably be based on DVB-SUB standards or standards based on W3C TTML, such as for example EBU-TT or SMPTE-TT (Society of Motion Picture and Television Engineers-Timed Text), or 30 Society of Animated Image and Television Engineers - Text Timed). US Patent 5497241 describes a system for displaying in a video block subtitles. However, the blocks are here predefined blocks, each block being associated with a language. Thus, the choice of language determines the shape of the blocks and the presentation. The subtitle coding standards and techniques available today therefore make it possible to represent subtitles with various styles (essentially fonts of different characters), to separate these subtitles into blocks and to display the subtitles. blocks of subtitles according to the desired style.
[0003] However, the known techniques of the state of the art do not allow the user to choose, from the same source of subtitles, a mode of presentation and segmentation in blocks. This problem especially affects live subtitles or the only way of presentation is often word-for-word. However, it also applies to pre-prepared captions, for which a pre-prepared format (e.g., line-by-line display) may be different from the user's preferred presentation mode. (for example, sentence by sentence). An object of the invention is therefore to propose a method for controlling the subtitle presentation form, making it possible to propose to the user several presentation modes for the same subtitle content, while leaving him great flexibility in choosing his preferred presentation mode. The invention can be applied regardless of the mode of transport. It can in particular apply to multimedia streams, for example MPEG-TS (Motion Picture Experts Group 20 - Transport Stream), or Expert Group of Motion Picture - Transport Stream, but also to files, for example files in ISOBMFF (International Standards Organization Base Media File Format) format, or International Standardization Organization's Basic Media File Format. It can also apply to any type of subtitle representation. It can in particular apply to subtitles represented as image sub-titles, for example in the context of the DVB-SUB standard, but also to subtitles represented in textual form, for example in the context of different standards based on W3C TTML (EBU-TT, SMPTE-TT, ..).
[0004] SUMMARY OF THE INVENTION [0013] To this end, the invention relates to a method for creating a subtitle stream or file composed of subtitle elements, said method comprising at least, for each subtitle element, a step of inserting said subtitle element into the stream or subtitle file; at least one step of checking at least one end-of-block condition, a block comprising at least one subtitle element, said end-of-block condition being linked to a subtitle presentation mode; when said at least one end-of-block condition is satisfied, inserting in the stream or subtitle file a data representative of the end of a block according to said subtitle presentation mode. In one set of embodiments of the invention, the end of block condition comprises the detection of at least one predefined character 15 marking the end of a block. [0015] Advantageously, at least one predefined character is a point, a comma, a semicolon or a punctuation character. In one set of embodiments, the invention applies to a method of creating a subtitle stream complying with the MPEG20 TS standard. In one set of embodiments, the invention applies to a process for creating a subtitle file respecting at least one standard based on W3C TTML. [0018] Advantageously, a datum representative of the end of a block is written in a specific tag. The invention also relates to a method of presenting a stream or a subtitle file on the terminal of a user, said method comprising at least one step of selecting on said terminal a mode of 30 presentation of subtitles; a step of reading a block of subtitles associated with said presentation mode in said stream or file; a presentation step on the terminal of said at least one block of subtitles according to said presentation mode. [0020] Advantageously, the method of presenting a stream or subtitle file 35 according to the invention comprises reading said stream or file 3025925 6 up to a datum representative of the end of a block of data. subtitles according to the selected subtitle presentation mode. In one set of embodiments, the invention applies to a method of presenting a subtitle stream complying with the MPEG5 TS standard. In one set of embodiments, the invention applies to a method of presentation of subtitle files respecting at least one standard based on W3C TTML. [0023] Advantageously, the data representative of the end of a block is written in a specific tag. The invention also relates to a system for creating a subtitle stream or file composed of subtitle elements, said system comprising at least, for each subtitle element, at least one configured module. to insert said subtitle element into the subtitle stream or file; at least one module configured to check at least one end-of-block condition, a block comprising at least one subtitle element, said end-of-block condition being related to a subtitle presentation mode; at least one module configured for, when said at least one end-of-block condition is satisfied, inserting into the stream or subtitle file a piece of data representative of the end of a block according to said presentation mode subtitles. The invention also relates to a user terminal configured to present a stream or file of subtitles to at least one user, said terminal comprising at least, for each subtitle element: at least one element to implement selects a subtitle presentation mode by said at least one user; at least one element configured to read a block of subtitles associated with said subtitle presentation mode; at least one element configured to present said at least one block of subtitles according to said at least one presentation mode. The invention also relates to a computer program for creating, when executed on a processor, a stream or subtitle file composed of subtitle elements, said computer program comprising at least for each subtitle element: computer code elements for inserting said subtitle element into the stream or subtitle file; computer code elements for checking at least one end-of-block condition, a block comprising at least one subtitle element, said end-of-block condition being related to a subtitle presentation mode; computer code elements for, when said at least one end-of-block condition is satisfied, inserting in the stream or subtitle file a datum representative of the end of a block according to said subtitle presentation mode . The invention also relates to a computer program configured, when it is executed on a processor, to present a stream or file of subtitles on the user terminal, said computer program comprising at least: computer code elements for implementing the selection of a subtitle presentation mode by at least one user; computer code elements for reading a block of subtitles associated with said subtitle presentation mode; computer code elements for presenting said at least one block of subtitles according to said at least one presentation mode. The method of the invention thus allows a user to choose the form of presentation he deems most appropriate for subtitles. The method according to the invention allows the user great flexibility in the presentation of subtitles. When implemented using metadata or extension of existing standards, the invention is backward compatible with existing subtitle decoding systems. [0031] An encoder can be updated with new presentation modes, transparently for the viewer who can choose from a wider choice of subtitle presentation modes. The invention is compatible with all modes of transmissions of subtitles, particularly in the form of streams or files. [0033] The invention is particularly applicable to multimedia streams according to MPEG-TS standards. The invention is particularly applicable to multimedia files according to ISOBMFF standards. The invention is applicable to subtitle representations in the form of images, for example in the DVB-SUB standard. The invention is applicable to representations of subtitles in the form of text, as for example for the various standards based on W3C TTML (EBU-TT, SMPTE-TT, ..).
[0005] LIST OF FIGURES [0037] Other characteristics will become apparent on reading the detailed description given by way of nonlimiting example, which follows with reference to the attached drawings which represent: FIGS. 1a and 1b, two examples of video encoders producing a stream with subtitles, one from live content, the other from a file, according to the state of the art; FIGS. 2a and 2b, diagrams illustrating multimedia flow structures integrating subtitles, respectively MPEG-TS and EBU-TT, one of the standards based on W3C TTML, according to the state of the art; FIG. 3, a flow diagram illustrating a process for creating subtitles according to the invention; FIGS. 4a, 4b, 4c and 4d, four examples of subtitles produced by a method according to the invention, respectively in an MPEG-TS stream structure and files according to the EBU-TT, SMPTE-TT and W3C TTML; FIG. 5, a flowchart illustrating a method of presenting a stream or subtitle file on a user's terminal; FIGS. 6a, 6b and 6c, three diagrams respectively illustrating three examples of subtitle presentation modes, according to the prior art in FIG. 6a and according to the invention in FIGS. 6b and 6c.
[0006] DETAILED DESCRIPTION [0038] In the rest of the description, the method according to the invention is illustrated by examples relating to the generation of subtitles in an encoder and the presentation of subtitles by a decoder, even if it should be noted that the invention can be applied to any type of device processing subtitles, including a transcoder included in a video processing chain. [0039] Certain English acronyms commonly used in the technical field of the present application may be used during the description and figures. These acronyms are listed in the table below, including their Anglo-Saxon expression and their meaning. Acronym Expression Meaning AAC Advanced Audio Coding Advanced Audio Encoding. Standard compression and audio coding. ABR Adaptive Bit Rate Adaptive Bit Rate. Describes a multimedia stream whose bit rate can be adapted, in particular to provide only sufficient data to achieve a given quality, or to adapt to a transmission capacity. AVC Advanced Video Advanced Video Encoding. Standard Coding compression and video coding, also called H.264. DDS Display Definition Definition Segment of the Display. Segment Segment of the DVBSUB standard indicating the resolution (Height x Width) on which the subtitles are to be displayed. DVB Digital Video Digital Video Broadcasting. Broadcasting Consortium has produced several standards for digital television. Also applies to produced standards, including the 302 5 92 5 10 DVBSUB subtitle standard, or ETSI EN 300 743 standard. DVBSUB Digital Video Subtitles Radio and Television Video Braodcasting Subtitles Digital. Closed captioning standard published by DVB, also known as ETSI EN 300 743. EBU European Broadcasting Union European Broadcasting Union. International Association of Broadcasters. The acronym also applies to standards produced by the organization. EBU-TT European Broadcasting Union-Timed Text European Broadcasting Union - Text Timed. EBU standard subtitle coding standard based on W3C TTML standard. ETSI European Institute for European Standards Telecommunication Standard Institute Telecommunication. European standardization body in the field of telecommunications. The acronym also applies to the standards produced by the institute. HD High Definition High Definition. Refers to a video stream of high resolution and quality, usually with a high bit rate per second. HEVC High Efficiency High Efficiency Video Coding. Standard Video Coding compression and video coding, also called H.265. ISO International Standard Organization Organization of International Standardization. Organization setting standards, especially in the multimedia field. ISOBMFF ISO Base Media Format of the Basic ISO Media File. Generic file format containing File Format multimedia data. Many file formats, such as the .mp4 or the 3025925 11 .m4a, are based on the ISOBMFF. MPEG Motion Picture Group of Experts of the Animated Picture. Experts Group Group of experts having published standards for multimedia content. This name is also applied to some of the product standards, for example the MPEG-2 MPEG-TS Motion Picture Expert Group Picture of Motion Picture - Transport Stream. Standard Defining an Experts Group Stream - Multimedia Transport Defined by the Transport Stream MPEG Committee. OTT Over The Top Above. Sending media from a content provider to a reader over a standard internet connection rather than a dedicated infrastructure. PCM Pulse Code Pulse Coding Modulation. Modulation Uncompressed digital representation of an analog signal via a sampling technique. PCR Program Clock Program Reference Clock. Reference Time data used to synchronize the different multimedia streams within a program in MPEG-TS streams. PES Packetized Elemental Stream by Package. Elementary Stream specification depending on the MPEG-2 standard, and describing the syntax of an elementary stream (video, audio, subtitles ...) within an MPEG-TS stream. A PES stream is divided into PES packets that can be multiplexed with packets of other PES elementary streams at the sin of an MPEGTS stream. PCS Composition page Segment of the DVBSUB standard indicating the composition of the subtitles page. Segment 3025925 12 PTS Presentation Time Time Stamp Presentation. Stamp Indicates a timestamp to which an element must be presented within a stream. SD Standard Definition Standard Definition. Designates a medium resolution video stream. SDI Serial Digital Digital Interface Series. Transport or broadcast interface protocol of different multimedia formats, in particular used as an internal communication mode between the different modules of a video encoder. SMPTE Society of Motion Engineers Society of Picture and Motion Picture and Television. International Television Engineers Association, located in the United States, developing video standards. SMPTE- TT SMPTE-Timed Text SMPTE Timed Text. Standard caption coding developed by SMPTE and based on W3C TTML. XML Extensible Markup Extensible Tag Language. Language Computer language for tagging information. W3C World Wide Web Consortium Consortium Web on Width of the World. Standards body responsible for promoting compatibility of web technologies, including XML. W3C W3C Timed Text Markup Language W3C Extensible Word Text Language. W3C TTML is a technique for presenting a textual support synchronously with another medium, for example audio or video. In the following description, we will call "subtitle element" a set of characters to be displayed on the screen simultaneously 3025925 13. A subtitle element usually consists of a word, a set of words, or a phrase. We will call "subtitle presentation mode" a set of rules for grouping and segmentation of subtitle elements in blocks of 5 subtitles to obtain a consistent presentation of subtitles. Thus, a subtitle mode defines how the subtitle elements are grouped together for presentation. For example, a phrase-by-sentence subtitle presentation mode of creating a subtitle block to be displayed simultaneously for each sentence in the subtitle stream can be considered. Thus the subtitle presentation mode describes how words and characters are grouped together, and should not be confused with the subtitle presentation style. The subtitle presentation style describes, for each character, group of characters, or for the entire stream or subtitle file, how it should be represented on the screen (font and font size) , color, etc ...). The TTML standard of the W3C, based on the XML format, allows a particularly flexible coding of the subtitles. It is designed to provide a generic and extensible subtitle description language, and there are particular implementations such as the EBU-TT standard and the SMPTE-TT standard. In the rest of the description, the terms "one of the W3C TTML standards", "one of the standards based on W3C TTML", "one of the W3C TTML standards", or similar expressions, will refer to the W3C TTML standard itself. and standards that are implementations thereof, such as the EBU-TT standard or the SMPTE-TT standard, for example. [0043] FIGS. 1a and 1b show two examples of video encoders producing a stream with subtitles, one from live content, the other from an input file, according to state of the art. FIG. 1a represents an example of a video encoder processing live subtitles known from the state of the art. This encoder notably comprises: a multimedia encoder 100a, itself comprising: a video encoder 101a; O An audio encoder 102a; O A subtitle encoder 103a; a flow multiplexer 104a; - A subtitle inserter 112a; SDI interfaces making it possible to transmit the multimedia streams within the encoder This encoder produces at least one live multimedia stream, for example solely illustrative and non-limiting of the scope of the invention, for television broadcasts of sporting events or weather reports. According to a current live state-of-the-art multimedia content design chain, a media stream 110a containing one or more audio and video channels is continuously produced. A subtitle stream 111a is then added in real time to the multimedia stream 110a by the subtitle driver 112a, to produce a raw multimedia stream 113a containing subtitles. These subtitles are generally produced by an operator retranscribing in real time the words expressed on the screen. In general, the operator writes each word separately, and then sends the word to the subtitle editor immediately to minimize the latency between the audio / video content and the subtitle content. The generated subtitle stream can thus, according to the variations of speed of writing and comprehension of the operator, present variable speeds of presentation of successive words, which, when the speed becomes excessive, may be detrimental to the quality subtitle produced. Indeed, this technique is likely to produce, on certain sequences, appearances of successive words extremely fast, which will thus be difficult to read by the viewer. The multimedia stream with subtitles 113a is then sent, for example via an SDI protocol, to the multimedia encoder 100a. The subtitle inserter can insert a single subtitle channel in the multimedia stream. It can also insert several subtitle channels in parallel, especially if the subtitles must be available in several languages. The subtitle inserter can insert subtitles according to all accepted standards within the SDI protocol. For example, it can insert teletext subtitles into the SDI. Within the multimedia encoder 100a, the different channels of a multimedia stream such as the stream 113a are separated in order to be processed by the corresponding encoding modules. Each video channel is processed by a video encoder 101a, each audio channel by an audio encoder 102a, and each subtitle channel by a subtitle encoder 103a. The video encoder 101a converts a video stream from the SDI into a compressed video format to allow transmission of the video stream with a reduced bit rate. The video stream within the SDI is coded according to the so-called YUV 4: 2: 2 mode. In this mode, the pixels are encoded in the YUV color space, in which the so-called "Y" layer contains the light intensity information, and the "U" and "V" layers contain the colorimetry information. In YUV 4: 2: 2 mode, some of the information contained in the U and V layers, considered less perceptible to the human eye, is suppressed. The video encoder 101a transforms this stream into a compressed stream with a much lower bit rate. In order to be easily readable, video streams are generally produced according to established standards, for example MPEG-2 Video, H.264-AVC, or H.265-HEVC standards. The audio encoder 102a converts an audio stream from the SDI compressed audio format to allow the transmission of the audio stream with a reduced bit rate. The audio stream within the SDI is usually coded in PCM.
[0007] For example, the PCM sampling values of the audio signal can be 48 kHz and 24 bits, which means that each audio channel comprises 48,000 samples per second, each of which contains a 24-bit coded loudness value. The audio encoder 102a transforms this stream into a compressed audio stream having a much lower bit rate. In order to be easy to read, audio streams are usually produced according to established standards, for example MPEG-3 Audio or MPEG-4-AAC standards. The subtitle encoder 103a converts a subtitle stream from the SDI into a stream of standard subtitles for broadcast. It converts in particular, from the formats supported by the SDI, the subtitles in a format that can be decoded by the standard decoders. The subtitles carried by the SDI may for example be teletext subtitles. By way of example, the subtitle encoder 103a can convert the subtitle channels carried by the SDI into subtitles according to the DVBSUB standard. It can also produce subtitles according to a standard using the XML format, in particular one of the standards based on W3C TTML. The stream multiplexer 104a concatenates the audio, video and subtitle streams coming from the encoders 101a, 102a and 103a within a multimedia stream 114a that can be sent to the user. A stream multiplexer can produce a single stream including audio, video and subtitle streams. It can also produce multiple media streams. For example, if audio and subtitles are available in multiple languages, it can produce a media stream for each of the languages used. Similarly, if the video encoder 101a is configured to produce multiple video streams at multiple compression levels and / or multiple resolutions, the stream multiplexer 114a can produce a multimedia stream at multiple levels of compression and / or resolution. . This is especially useful when the feed is to be sent to clients with different resolution and / or bandwidth contexts. For example, a stream multiplexer 104a can prepare both a high quality HD stream for cable connected televisions, and an SD stream for mobile devices. The stream 114a product may in particular be of the MPEG-TS type for a live transmission. An MPEG-TS stream may contain video, audio and subtitle channels, as well as timing and description information of available channels. It can be produced and sent continuously by the multimedia encoder 100a. An MPEG-TS stream allows so-called multicast broadcasting, ie for which the encoder produces and broadcasts a single stream, which is received by different receivers. This type of stream is for example commonly used in terrestrial digital television systems. The multimedia product stream may also be of the ISOBMFF (ISO Based Media File Format) file format. This type of file is commonly used for OTT video transmissions. The OTT acronym stands for Over The Top, that is above the entire network.
[0008] The purpose of this mode of transmission is to be able to transmit a multimedia stream via any internet connection, without a dedicated network. This type of stream consists in sequencing the multimedia stream into short files, for example of the order of a few seconds, for example between 1 and 10 seconds. The files can then be sent as soon as they are trained to a client or reader, who will read them as and when. In this type of stream, a manifest file describing the different streams available must be sent beforehand to the multimedia files. This manifest file can contain, for example, the list of audio and subtitle channels, if these are available in several languages. This type of transmission also allows so-called ABR modes for Adaptive Bit Rate. In these modes, the video encoder produces several video streams, at several bit rates, and the stream multiplexer 104a produces small separate files at these different bit rates. Depending on the available bandwidth and the requests sent by the clients, files with more or less important bit rates are sent, in order to adapt the bit rate of the multimedia stream to the transmission capacity for each client. FIG. 1b represents an example of a multimedia encoder operating from a file according to the state of the art. The encoder 100b takes as input a multimedia file 113b containing video, audio and subtitle channels. It separates the different channels and provides them to video encoders 101b, audio 102b and subtitles 103b similar to encoders 101a, 102a and 103a. Encoders 101b and 102b may also include decompression means when the audio / video streams of file 113b are compressed. The streams are then decompressed and then re-compressed by the encoders. In this case, we are talking about transcoding. It is also possible, when the streams present in a file 113b are compressed, not to modify one of them, for example by preserving the initial audio stream, while transcoding the video stream. The subtitle encoder 103b can also transcode the subtitles, for example by transforming teletext subtitles and subtitles according to one of the TTML standards of the W3C. It can also make changes to the subtitles themselves, for example changing their fonts or timings. A multiplexer 104b according to the state of the art is capable of multiplexing the different audio, video and subtitle streams within one or more multimedia streams 114b. Multiplexers 104b and multimedia stream 114b respectively have the same properties as multiplexers 104a and multimedia stream 114a. [0056] FIGS. 2a and 2b show diagrams illustrating multimedia flow structures integrating subtitles, respectively MPEG-TS and EBU-TT, one of the standards based on W3C TTML, depending on the state of the art. These flows may be for example flows 114a or 114b produced by flow encoders 104a or 104b. The invention can be applied to these types of streams, but also to any type of stream incorporating timed subtitles. FIG. 2a represents the structure of a packet 20a of MPEG-TS stream that can contain subtitles according to the state of the art. This packet 20a may contain, for example, audio, video or subtitles, and comprises: A header 200a; Useful information 210a, comprising: a packet start prefix 211a; o A stream identifier 212a; O A packet size 213a; o Two so-called "flags" bytes making it possible to apply properties to the packet 214a; o Packet size information 215a; o A presentation timestamp, or PTS 216a; O Various other header fields 217a; o A range of useful data 218a; The header 200a includes synchronization information with the entire stream. The useful information 210a includes all the information to correctly present the packet 20a. The stream identifier 212a allows a reader to determine which stream belongs to the packet. The different streams may in particular be numbered 1,2, 3, etc. This identifier refers to description information sent to the stream at regular intervals. This description information notably lists all the streams, as well as the data they contain. Thus a stream may designate an audio, video or subtitle stream whose properties have been defined previously. In particular, the description of the subtitle stream describes the type of subtitle stream, for example if the stream is a DVBSUB type subtitle stream. The decoder is thus capable, knowing the type of stream, to decode it via a decoder 35 or appropriate decoding algorithm. The packet size information 215a contains the size of the packet. It allows a decoder to retrieve all the useful information of the packet, in order to decode it correctly. Presentation timestamp 216a contains the timestamp 5 to which the subtitle contained in the packet is to be displayed. This timestamp is expressed in a common reference to audio, video and subtitle packets, which allows these three data to be displayed synchronously. The unit of the time stamp is standardized (the PTS are reduced to a time unit of 90000 samples per second), and the initial time stamp is defined by the service clock, which corresponds to the time stamp. beginning of presentation of the program in the decoder. A decoder is therefore able, from these data, to display the subtitles synchronously with the audio and video. The useful data range 218a contains the raw data related to the subtitle element to be presented. For example, if the subtitle stream complies with the DVBSUB standard, it may be a bitmap image according to the commonly accepted name in this technical field. In this case, each subtitle element is represented as an image superimposed on the video. For each image, the luminous intensity values of each pixel are stored one after the other to form an image representative of the characters to be transmitted in the subtitle. FIG. 2b shows a subtitle sequence 200b according to the EBU-TT standard, one of the standards based on W3C TTML, according to the state of the art. A subtitle according to this standard is organized hierarchically with different levels of XML tags. The subtitle sequence 200b notably comprises a header 210b. This header contains information on the different styles (fonts, font sizes, colors, alignments ...) to apply to the subtitle. It may also contain information on the language of subtitle 200b, "en" meaning English in this case. The subtitle sequence 200b also contains, for each element, a body 220b or "body" section, containing at least one 3025925 section called "div" 230b. For each subtitle element, a section "p" 240b contains in particular an identifier number 243b, a start timestamp 241b and an end timestamp 242b. The start and end time stamps are expressed in absolute time.
[0009] The player can thus synchronize the presentation of the subtitle element with audio and video. A span tag 250b includes the text to be displayed 251b, and may contain 252b style information. In a state-of-the-art encoder 100a or 100b, a subtitle element is created for each subtitle element of the incoming stream or file, for example elements of the stream 113a or the file. 113b. When these subtitles are produced live, these subtitle elements can be presented word for word, and lead to an unpleasant user experience. The invention describes a method of controlling the subtitle presentation form. It can therefore be applied to any system, method or computer program processing subtitle sequences comprising at least two elements. For example, this invention can be integrated into subtitle encoders 103a or 103b. The invention can be applied to any subtitle format. For example, it can be applied to MPEG-TS 20a subtitle packets, or subtitle sequences complying with one of the W3C200b TTML standards. It should be noted, however, that these examples are for illustrative purposes only. Thus, the invention also applies to a decoder, which displays on the terminal of a user the subtitles according to the desired presentation mode. It may also apply to a transcoder, which will generate, from a first subtitle sequence, a second sequence of which each element will be derived from a block of subtitles according to a presentation mode selected within of the first sequence. FIG. 3 represents a diagram illustrating a method according to the invention. A method according to the invention applies to a subtitle element 310 comprising in particular: a step of inserting the subtitle element into a stream or subtitle file 320; A check step 330 of at least one end-of-block condition 340, said end-of-block condition being related to a subtitle presentation mode 360; When at least one end-of-block condition is verified, the insertion in the subtitle stream of a data representative of the end of a block according to said presentation mode of subtitles 350; A subtitle element 310 may consist of a word, a set of words or more generally a sequence of characters originally intended to be displayed at the same time stamp. For example, it may be words contained in PES 218a packet payload bytes, or 252b words contained within a span of a file according to one of the W3C TTML based standards. The subtitle element may be derived from a stream, a file, or more generally from any source for displaying a stream of subtitles comprising successive elements. Step 320 of inserting the subtitle element into a stream or subtitle file is to insert the contents of the subtitle element 310 into the stream or output file. . The insertion is specific to each type of file / output stream, and may include writing data to a computer readable memory, or sending data over a network. Step 320 may in particular comprise creating, in the output stream or file, a subtitle element comprising the data contained in the element 310. It may for example include, if the stream or output file is an MPEG-TS stream, creating at least one MPEG-TS packet such as that shown in Figure 2a; if the stream or output file is a file according to one of the W3C TTML based standards, creating and writing at least one "div" tag and a "span" tag. The step 330 of checking at least one end-of-block condition 340 is based on the content of the subtitle element 310, a set of subtitle presentation modes 360 comprising end of block 340, and possibly on the history of previously inserted elements in the stream or output file. The presentation modes of 360 subtitles are a set of subtitle presentation modes, such as, by way of non-limiting examples, "line by line", "sentence by sentence", "by group at least 302 5 9 2 5 22 50 characters ", etc. These presentation modes may correspond to user preferences. Each presentation mode is associated with an end-of-block condition 340. An end-of-block condition is a rule for separating the subtitle blocks for a given presentation mode, and generally follows from the definition. of this mode of presentation. For example, for a "sentence-by-sentence" presentation, the end-of-block condition will be the presence, within the subtitle element, of a representative element of an end of sentence, for example a point. Similarly, for a presentation mode "per group of at least 50 characters", a block will be formed when the number of characters accumulated by the different subtitle elements since the last block will be greater than or equal to 50. another example, for an "end of time" presentation mode, it is expected that a block is displayed as soon as a duration is exceeded between two successive blocks. In this presentation mode, as soon as a block is created, its presentation timestamp is noted, and the next block is created as soon as a subtitle element whose timestamp is after the timestamp presentation time of the first block, to which is added the minimum duration, is identified. These examples are given in a nonlimiting manner, and a wide variety of presentation modes and block segmentation rules are a priori permissible. Step 330 then consists in determining what are the end-of-block conditions 340 verified by the subtitle element 310. When at least one end-of-block condition 340 is checked by 25 subtitle element 310, the step 350 of inserting a data representative of an end of block is activated. This step consists of writing to a file or inserting into a stream a datum expressing the end of the block of subtitles for the presentation mode concerned. Examples will be given in the rest of the description, in examples of MPEG-TS 30 streams and files according to one of the standards based on W3C TTML. In one embodiment of the invention, each of the end-of-block conditions 340 is verified for each of the 360 subtitle presentation modes; whenever a condition 340 is verified, step 350 of inserting a data representative of a block end 35 is enabled for the corresponding subtitle presentation mode 360. FIGS. 4a, 4b, 4c and 4d show four examples of subtitles produced by a method according to the invention, respectively in an MPEG-TS stream structure and in files according to the EBU-TT standards. , 5 SMPTE-TT and W3C TTML. FIG. 4a represents an exemplary MPEG-TS 400a packet produced by a method according to the invention, in an MPEG-TS stream structure. This MPEG-TS packet comprises, as a state-of-the-art MPEG-TS packet, a header 200a, a payload 210a, a packet start prefix 211a, a stream identifier 212a, and a packet end byte 218a. It may also include a DDS segment 401a indicating the resolution on which to display the subtitles, and a PCS segment 402a indicating the composition of a subtitle page such as the number of regions to be displayed and their positions in the page. . [0080] 11 also comprises a metadata segment 410a, comprising a synchronization byte 411a, a segment type 412a, a page identifier 413a, a segment length 414a and at least one byte representative of a block end 420a. Synchronization bytes 411a, segment type 412a, page identifier 413a, and segment length 414a are known from the state of the art. This set of elements makes it possible to define a new type of segment, indicating the ends of blocks. The at least one byte representative of a block end 420a is created during the step of inserting an end-of-block representative data 350, and comprises at least one byte representative of the end of at least one block according to a presentation mode. By way of non-limiting example, the packet 400a comprises bytes representative of the end of 5 blocks according to 5 different presentation modes: "end of line" 421a, "end of time" 422a, "new color" 423a , "Simple grammar analysis" 424a and "improved grammar analysis" 425a. In the case in point and as a non-limitative example, the "end of line" mode generates a block for each line of subtitles; the "end of time" mode generates a block for all the subtitle elements generated during successive periods of time whose duration is fixed; "new color" mode generates a block at each color change in DVBSUB subtitles; the "simple grammar analysis" mode generates a block each time it encounters certain punctuation characters, for example a point; finally, the "enhanced grammar analysis" mode generates a block for each punctuation mark limiting sentences or part of sentences. The "enhanced grammar analysis" mode then stands out from the "simple grammar analysis" mode in that, when encountering a punctuation character, it causes a semantic analysis to determine whether this punctuation character delimits sentences or propositions. For example, in the presence of a comma, the "enhanced grammar analysis" mode will generate a new block only if this comma separates two propositions within a sentence. The "new color" mode integrates a block end with each color change. This mode is interesting in the case of some pre-processed subtitles in which each color has a particular meaning. For example, some subtitles assign a color to each character. The mode of implementation "new color" then makes it possible to automatically segment the subtitles by block according to the character having the speech. In an implementation in the DVBSUB standard, the step of checking the end-of-block condition may include analyzing, within sub-title images in the DVBSUB standard, dominant RGB color triplets. These examples of presentation modes are given as an illustration of the wide field of application of the invention. Other examples may be introduced without departing from the scope of the invention. Many implementations are possible for creating end-of-block bytes 350. In one embodiment of the invention, at least one byte is allocated for each presentation mode, the content indicates whether a block end is reached for this presentation mode or not. In this embodiment, the number of bytes in the set of bytes representative of the block ends 420a remains constant. In another embodiment of the invention, a metadata segment 410a containing bytes representative of the ends of blocks 420a is created if and only if an end of block is reached for at least one presentation mode. The bytes representative of an end of block according to a presentation mode, for example "end of line" 421a, contain an identifier designating this mode. A block is terminated for each of the modes whose identifier is present in the at least one byte representative of an end of block 420a. In a third embodiment of the invention, a metadata segment 410a is created for each end of block in a presentation mode. The at least one octet representative of a block end 420a then contains only a presentation mode identifier, for example "end of time" 422a. For a packet 400a, as many metadata segments 410a are created as end of blocks are encountered for the subtitle element contained in the packet 400a. These modes of implementation are described by way of non-limiting examples. In a set of embodiments of the invention, a header sent periodically in the MPEG-TS stream contains description information for at least one subtitle presentation mode. This information may include the name and description of the presentation mode. FIG. 4b represents an example of a 400b subtitle file produced by a method according to the invention, according to the EBU-TT standard, one of the W3C TTML standards. This file 400b comprises, as a file according to the standard EBU-TT according to the state of the art, a 410b header, a body contained in a body section 420b, at least a div section 430b, comprising at least one "p" section 440b. As in a state-of-the-art EBU-TT standard file, the "p" section 440b includes at least one start time stamp 441b, an end time stamp 442b, a flow identifier 443b, a span section 450b, text to display 451b and may include style information 452b. In a set of embodiments of the invention, the "p" tag 440b of this file 400b also comprises at least one datum representative of the end of a block according to a presentation mode, for example a beacon "Tt: display mode" 460b. In one embodiment of the invention, this tag contains a reference on each of the available modes, indicating whether the subtitle element included in the tag "p" 440b corresponds to a block end, for each of these modes. By way of nonlimiting example, the beacon 460b of FIG. 4b indicates a end of block 3025925 26 according to three modes: "end of line" 461b, "Simple grammar analysis" 464b and "improved grammar analysis" 465b, whereas it indicates a lack of end of block for two modes "End of time" 462b and "New color" 463b. In the example shown, the presence or absence of an end of block is indicated respectively by the keyword "true" or "false". In another embodiment, it may be indicated by predefined values such as "1" and "0". It can also be indicated by the words "true" or "false", or in general by any keyword expressing whether a proposition is true or not. [0086] The flexibility of the XML languages allows a wide variety of embodiments of the invention to indicate an end of block in the EBU-TT standard, and more generally in W3C TTML based standards. By way of nonlimiting example, in one embodiment, the beacon 460b comprises, instead of the exhaustive list of modes, a list of modes for which the current subtitle element corresponds to a block end. In a second embodiment, a tag is created for each end of the block according to a presentation mode. In a third embodiment, a list of presentation modes for which the current subtitle element is an end of block is integrated as an attribute of a tag, for example the "span" tag 450. In this mode realization, the tag 450b could become for example: <tt: span style = "WhiteOnBlack '" list_end_blocs = "end_of_line mg_pars_g ram may simple_gram_gram">, indicating that the current subtitle element marks an end of block for the three "end of line" presentation modes , "Simple grammar analysis" and "Improved grammar analysis". Of course, this list can also be included in another tag, for example the tag "p" 440b or "div" 430b. In a set of embodiments of the invention, the header 410b contains description information for at least one subtitle presentation mode. This information may include the name and description of the presentation mode. , as well as defining a presentation style (font, character color) to be assigned to said presentation mode for optimal display. FIGS. 4c and 4d show an example of a subtitle file produced by a method according to the invention, respectively according to the SMPTE-TT and W3C TTML standards. Since the SMPTE-TT and EBU-TT standards are derived from the W3C TTML standard, the 400c and 400d files produced by a method according to the invention according to the SMPTE-TT and W3C TTML standards are similar to a file 400b produced according to the EBU standard. -TT. Thus, the headers 410c, 410d, body 420c, 420d, tag "div" 430c, 430d tags "p" 440c, 440d, start timestamp 441c, 441d, end timestamp 442c, 442d, identifier flow 443d, tag 10 "span" 450c, 450d, display text 451c, 451d are they respectively similar, apart from some semantic differences between the standards, 210b headers, 410b, body 220b, 420b, tag " div »230b, 430b," p "tags 240b, 440b, start time stamps 241b, 441b, end time stamp 242b, 442b, stream identifier 243b, 443b, span tag 250b, 450b, and display text 251b, 451b in the EBU-TT standards according to the state of the art and according to the invention. The semantic differences between the standards integrate, in particular, different tag naming, for example the header tag, named "tt: head" in the EBU-TT standard, and "head" in the 20 W3C TTML standards. and SMPTE-TT. W3C TTML-dependent standards may also contain additional information and tags, such as ttm: title, ttm: desc, ttm: copyright, smpte: information contained in the 410c header tag, and information original 444c in the SMPTE-TT standard. [0091] However, these differences do not affect the ability to implement the invention in the various standards derived from W3C TTML. Thus, one skilled in the art can easily implement the invention in each of the W3C TTML standards in an equivalent manner to its implementation in EBU-TT. By way of nonlimiting example, the "display mode" tags 460c and 460d contain end-of-block rules identical to those of the tag 460b, with a value representative of the presence or absence of a block end, for example. example according to end-of-line presentation modes 461c, 461d, end-time 462c, 462d, new color 463c, 463d, simple grammar analysis 464c, 464d, improved grammar analysis 35 465c, 465d . In the same way as for the EBU-TT standard, the flexibility of XML languages makes it possible to implement the invention in various ways in standards based on W3C TTML. By way of nonlimiting examples, it is possible to insert a tag for each end of block in the tags "p" 440c, 440d, or to insert a list of blocks whose current subtitle element marks the end within the "span" tags 450c, 450d or "p" 440c, 440d. In a set of embodiments of the invention, the headers 410c, 410d contain description information for at least one subtitle presentation mode. This information may include the name and description of the mode of presentation. presentation, as well as the definition of a presentation style (font, character color) to be assigned to said presentation mode for optimal display. [0094] FIG. 5 represents a flow diagram illustrating a method of presenting a stream or a subtitle file on the terminal of a user. This process can be implemented for the presentation of subtitles on the terminal of a user, for example in the context of the presentation of a video with subtitles. It may for example be implemented in a set-top box, a DVD / Blu-ray player, a computer or smartphone having multimedia functions. This process comprises in particular: a step 510 of selection on the terminal of a presentation mode of subtitles; A step 520 of reading a block of subtitles associated with said presentation mode in said stream or file; A step 530 of presentation on the terminal of said at least one block of subtitles according to said presentation mode; [0096] Step 510 consists of selecting the subtitle presentation mode on the terminal. This choice can be made manually, depending on the capabilities of the terminal. For example, it can be done using a menu and a remote control on a set-top box, or via a touch interface on a smartphone. The list of presentation modes may include only the names of the modes. It can also be associated with a description and / or graphical representation of each of these modes. The choice of the presentation mode can also be automatic 302 5 9 2 5 29. In this case, the terminal can, if it has a user preference file, automatically select the preferred mode of a user. In the context of video applications connected between different devices, for example between a laptop and a set-top box, a terminal can also automatically select the last mode used by the user, even if it were a subtitle display on a different terminal. It is also possible to provide on the same terminal a mode for each user, said mode being automatically selected when this user connects to the terminal. Finally, it is possible for the same user to have several modes on a terminal, and to select the one that seems most appropriate to him, for example when starting up the terminal or at the beginning of playing a video. Step 520 consists in reading a block of subtitles in the stream or file. It may include reading one of at least a portion of a subtitle file or stream. In a set of implementation modes of the invention, it comprises reading the characters to display until meeting a data representative of the end of a block of subtitles according to the selected presentation mode. Step 530 includes the presentation of the subtitle block read at step 530, depending on the capabilities of the terminal. In a set of modes of implementation of the invention, the block segmentation itself makes it possible to display the block of subtitles in the desired manner. This is for example the case for a display "sentence by sentence", for which a block forms a sentence. Other modes of presentation according to other modes of implementation of the invention constrain the presentation of the subtitles. For example, in a "line-by-line" presentation mode, each block must be displayed on a new line within the terminal. In a set of embodiments of the invention, the step 520 of reading a block of subtitles comprises at least: A step 521 for reading a subtitle element; A step 522 of adding the element to the current block of subtitles; A step 523 for checking the end of the block according to the mode selected [00100] The step 521 comprises reading a subtitle element in a stream or file. It may correspond to the reading of a TS 35 400a packet in an embodiment of the invention according to the MPEG-TS standard, or the reading of a subtitle tag 440b, 440c, 440d in one embodiment of the invention according to a standard W3C TTML. Step 522 includes adding the content of the subtitle element to a current block of captions. The content of the subtitle element may comprise characters, for example those contained in payload 217a of a TS packet or characters 251b contained in a tt: span tag. Step 523 consists in verifying, within the read subtitle element, the presence or absence of an element representative of an end of block 10 according to the selected presentation mode. In the example represented in FIG. 4a, this step comprises checking, in the byte of the range 420a linked to the selected presentation mode, whether it indicates an end of block or not. In the example represented in FIG. 4b, this step comprises the checking, in the tag tt: display 460b, of the presence or absence of end of block 15 for the selected presentation mode. If a block end is detected, step 530 block presentation is activated, and the current block reset. In the opposite case, a new subtitle element is read in step 521. [00104] In one set of embodiments of the invention, the subtitles are represented in textual form. This is for example the case in subtitles 400b, 400c and 400d, with a textual representation of subtitle elements 451b, 451c and 451d. [00105] FIGS. 6a, 6b and 6c show three diagrams respectively illustrating three examples of subtitle presentation modes, according to the prior art in FIG. 6a and according to the invention in FIGS. 6b and 6c [00106] FIG. 6a represents an example of subtitle presentation mode according to the prior art. It includes a time axis 600 indicating the moment when the subtitle elements are displayed, a line 610 representing the subtitles transmitted and a line 630a representing the subtitles displayed according to the state of the art. In this example, the transmitted subtitles comprise 10 elements 611, 612, 613, 614, 615, 616, 617, 618, 619, 620 and 621. These subtitles can be transmitted in the form of streams, for example MPEG-TS stream. In this case, each subtitle element 3025925 corresponds to at least one TS packet 20a, whose presentation time stamp is a PTS 216a. These subtitles can also be transmitted as a file, for example files according to one of the standards based on W3C TTML. In this case, each subtitle element comprises at least one "p" tag 440b, 440c, 440d, with a presentation timestamp 441b, 441c, 441d. [00107] A display according to the state of the art updates the presentation of subtitles for each new element, for example by concatenating it to the previous elements. Thus 10 successive displays 631a, 632a, 633a, 634a, 635a, 636a, 637a, 638a, 639a, 640a and 641a are obtained. Each of these subtitle displays is performed at the timestamp defined by the corresponding subtitle element. The absence of unity in the presentation form of the subtitles can then be unpleasant for the user. FIG. 6b shows an example of presentation of subtitles according to the invention. It includes the time axis 600, the line of the subtitles transmitted 610, identical to that of Figure 6a. It also includes a line 630b representing the successive presentation of the subtitles in this example, and a line 650b representing the end-of-block representative data according to an "end of line" presentation mode. By way of non-limiting example, the mode selected for display on the user's terminal is the "end of line" mode. The subtitle transmission includes, in addition to elements 611 to 621, end-of-block representative data, at least the end-of-line pattern end-of-line data 656b and 661b in this example. If the subtitles are transmitted in an MPEG-TS stream, the data 656b and 661b may for example be included in the metadata 421a. If the subtitles are transmitted in a file according to one of the W3C TTML based standards, the data 656b and 661b may for example be included in a tag 460b, 460c or 460d. During the presentation on the user's terminal, a block 636b, 641b comprising the content of the previous subtitle elements is displayed at the time stamp of the representative end-of-block data item 656b, corresponding 661b. A proper positioning of the end-of-block representative data to the encoding thus makes it possible to display the sub-titles line by line and to obtain a more pleasant display of the subtitles for the user. [00111] Figure 6c shows an example of subtitle presentation according to the invention. It comprises the time axis 600, the transmitted subtitle line 610, identical to that of FIG. 6a. It also includes a line 630c representing the successive presentation of the subtitles in this example, and a line 650c representing the end-of-block representative data according to a "simple grammar" presentation mode. By way of nonlimiting example, the mode selected for display on the user's terminal is here the "simple grammar" mode. Analogously to the preceding example, the transmission of the subtitles includes, in addition to the elements 611 to 621, end-of-block representative data, at least the data representing block endings according to the "simple grammar" mode. 651c and 658c in this example. During the presentation on the user's terminal, a block 631c, 638c comprising the content of the previous subtitle elements is displayed at the time stamp of the representative end-of-block data item 656b, corresponding 661b. In this example, a proper positioning of the end-of-block representative elements makes it possible to display the subtitles until meeting a punctuation mark such as a comma or a point, and to obtain a display of the subtitles. more pleasant for the user. [00114] The examples given above are given for information only, and in no way limit the scope of the invention, defined by the claims below.
权利要求:
Claims (15)
[0001]
REVENDICATIONS1. A method of creating a subtitle stream or file composed of subtitle elements, said method comprising at least, for each subtitle element (310): a step of inserting said subtitle element; titles in the stream or subtitle file (320); At least one verification step (330) of at least one end-of-block condition (340), a block comprising at least one subtitle element, said end-of-block condition being linked to a presentation mode of the subtitles (360); When said at least one end-of-block condition is verified, inserting in the stream or subtitle file a data representative of the end of a block according to said subtitle presentation mode (350) ;
[0002]
2. Method according to claim 1, characterized in that the end of block condition comprises the detection of at least one predefined character marking the end of a block.
[0003]
3. Method according to claim 2, characterized in that at least one predefined character is a point, a comma, a semicolon or a punctuation character.
[0004]
4. A method of creating a subtitle stream according to one of claims 1 to 3, characterized in that the subtitle stream complies with the standard MPEG-TS (400a).
[0005]
5. A method of creating a subtitle file according to one of claims 1 to 3, characterized in that the file complies with at least one standard based on W3C TTML (400b, 400c, 400d). 3025925 34
[0006]
6. Method according to claim 5, characterized in that a data representative of the end of a block is written in a specific tag. (460b, 460c, 460d). 5
[0007]
7. A method of presenting a stream or subtitle file on a user's terminal, said method comprising at least: a step of selecting on said terminal a subtitle presentation mode (510); A step of reading a block of subtitles associated with said presentation mode in said stream or file (520); A presentation step on the terminal of said at least one block of subtitles according to said presentation mode (530);
[0008]
8. The method as claimed in claim 7, characterized in that the step of reading a block of subtitles comprises reading said stream or file to a piece of data representative of the end of a block of subtitles. according to the selected subtitle presentation mode.
[0009]
9. The method of presenting a subtitle stream according to one of claims 7 and 8, characterized in that the stream complies with the MPEG-TS standard.
[0010]
10. The method of presenting a subtitle file according to one of claims 7 and 8, characterized in that the file complies with at least one standard based on W3C TTML (400b, 400c, 400d).
[0011]
11. A method of presenting a subtitle file according to claim 10, characterized in that a data representative of the end of a block is written in a specific tag. (460b, 460c, 460d). 30
[0012]
12.System for creating a stream or subtitle file composed of subtitle elements, said device comprising at least, for each subtitle element (310): - at least one module configured to insert said subtitle element subtitle element in the stream or subtitle file (320); At least one module configured to check at least one end-of-block condition (340), a block comprising at least one subtitle element, said end-of-block condition being linked to a sub-presentation mode. securities (360); At least one module configured for, when said at least one end-of-block condition is verified, inserting in the stream or subtitle file a datum representative of the end of a block according to said sub-presentation mode; titles (350). 10
[0013]
User terminal configured to present a stream or subtitle file to at least one user, said terminal comprising at least: at least one element configured to implement the selection of a subtitle presentation mode by said at least one user (510); At least one element configured to read a block of subtitles associated with said subtitle presentation mode (520); at least one element configured to present said at least one block of subtitles according to said at least one presentation mode (530); 20
[0014]
14. Computer program for creating, when executed on a processor, a stream or subtitle file composed of subtitle elements, said computer program comprising at least, for each sub-item -titles: - pieces of computer code for inserting said subtitle element in the stream or subtitle file (320); - Computer code elements for checking at least one end-of-block condition (340), a block comprising at least one subtitle element, said end-of-block condition being related to a subtitle presentation mode (360); Elements of computer code for, when said at least one end-of-block condition is satisfied, inserting in the stream or subtitle file a datum representative of the end of a block according to said sub-presentation mode; titles (350). 3025925 36
[0015]
15. A computer program configured for, when run on a processor, presenting a stream or subtitle file on the user terminal, said computer program comprising at least: - Computer code elements to implement effecting the selection of a subtitle presentation mode by at least one user (510); - computer code elements for reading a block of subtitles associated with said subtitle presentation mode (520); Computer code elements for presenting said at least one subtitle block according to said at least one presentation mode (530).
类似技术:
公开号 | 公开日 | 专利标题
US10950270B2|2021-03-16|Audio modification for adjustable playback rate
US9854375B2|2017-12-26|Selection of coded next generation audio data for transport
US8229748B2|2012-07-24|Methods and apparatus to present a video program to a visually impaired person
CN103988520B|2017-05-10|Reception device, method for controlling same, distribution device, distribution method and distribution system
US20210014574A1|2021-01-14|Using Text Data in Content Presentation and Content Search
FR3025925A1|2016-03-18|METHOD FOR CONTROLLING PRESENTATION MODES OF SUBTITLES
US20130204605A1|2013-08-08|System for translating spoken language into sign language for the deaf
KR101802273B1|2017-11-28|Apparatus and method for transmitting/receiving contents cosisting of a pluarlity of streams
FR2850821A1|2004-08-06|Audio signal e.g. television signal, sub-titling system for e.g. deaf and dumb people, has combining unit combining delayed audio signal and subtitling signal into subtitled audio signal applied to receiver equipment
US10750149B2|2020-08-18|Information processing apparatus, information recording medium, information processing method, and program
US10299009B2|2019-05-21|Controlling speed of the display of sub-titles
JP2021122139A|2021-08-26|Receiver and program
JP2016082473A|2016-05-16|Video playback device and video playback method
WO2017191397A1|2017-11-09|Method and device for synchronising subtitles
JP2008294722A|2008-12-04|Motion picture reproducing apparatus and motion picture reproducing method
EP1343327B1|2011-06-15|Method for processing multimedia content
JP2020178262A|2020-10-29|Live subtitle shaping processing device and program
EP3866106A1|2021-08-18|Broadcast signal transmission method, broadcast signal transmission device, broadcast signal reception method, and broadcast signal reception device
KR100849492B1|2008-07-31|Apparatus and Method for Converting Format of Digital Content including Character in Video
FR3004052A1|2014-10-03|METHOD AND DEVICE FOR TRANSMITTING VIDEO CONTENT, METHOD AND DEVICE FOR RESTITUTING VIDEO CONTENT, VIDEO FLOW AND CORRESPONDING COMPUTER PROGRAM.
WO2012080651A1|2012-06-21|Enrichment of the audio content of an audiovisual program by means of speech synthesis
KR20120029928A|2012-03-27|Method and apparatus for representing attribute of content component in a streaming service
同族专利:
公开号 | 公开日
FR3025925B1|2016-12-23|
US10341631B2|2019-07-02|
WO2016041859A1|2016-03-24|
US20170302900A1|2017-10-19|
引用文献:
公开号 | 申请日 | 公开日 | 申请人 | 专利标题
US5572260A|1995-03-20|1996-11-05|Mitsubishi Electric Semiconductor Software Co. Ltd.|Closed caption decoder having pause function suitable for learning language|
US8782721B1|2013-04-05|2014-07-15|Wowza Media Systems, LLC|Closed captions for live streams|
US20080085099A1|2006-10-04|2008-04-10|Herve Guihot|Media player apparatus and method thereof|
US20110013888A1|2009-06-18|2011-01-20|Taiji Sasaki|Information recording medium and playback device for playing back 3d images|
US20150215564A1|2014-01-30|2015-07-30|Echostar Uk Holdings Limited|Methods and apparatus for creation of a reference time index for audio/video programming|CN107251560B|2015-02-23|2021-02-05|索尼公司|Transmission device, transmission method, reception device, reception method, information processing device, and information processing method|
CN109155866A|2016-05-31|2019-01-04|索尼公司|Sending device, sending method, reception device and method of reseptance|
GB2553785A|2016-09-13|2018-03-21|Sony Corp|A decoder, encoder, computer program and method|
WO2021102355A1|2019-11-21|2021-05-27|Vooks, Inc.|Systems and methods for enhanced video books|
法律状态:
2015-08-25| PLFP| Fee payment|Year of fee payment: 2 |
2016-03-18| PLSC| Search report ready|Effective date: 20160318 |
2016-07-08| TP| Transmission of property|Owner name: HARMONIC INC., US Effective date: 20160607 |
2016-08-26| PLFP| Fee payment|Year of fee payment: 3 |
2017-08-23| PLFP| Fee payment|Year of fee payment: 4 |
2018-08-17| PLFP| Fee payment|Year of fee payment: 5 |
2019-08-19| PLFP| Fee payment|Year of fee payment: 6 |
2020-08-14| PLFP| Fee payment|Year of fee payment: 7 |
2021-08-31| PLFP| Fee payment|Year of fee payment: 8 |
优先权:
申请号 | 申请日 | 专利标题
FR1458793A|FR3025925B1|2014-09-17|2014-09-17|METHOD FOR CONTROLLING PRESENTATION MODES OF SUBTITLES|FR1458793A| FR3025925B1|2014-09-17|2014-09-17|METHOD FOR CONTROLLING PRESENTATION MODES OF SUBTITLES|
US15/510,149| US10341631B2|2014-09-17|2015-09-11|Controlling modes of sub-title presentation|
PCT/EP2015/070829| WO2016041859A1|2014-09-17|2015-09-11|Method of controlling modes of presentation of sub-titles|
[返回顶部]